How does TensorFlow work?

TensorFlow defines computations as Graphs, and these are made with operations (also know as “ops”). So, when we work with TensorFlow, it is the same as defining a series of operations in a Graph.

To execute these operations as computations, we must launch the Graph into a Session. The session translates and passes the operations represented into the graphs to the device you want to execute them on, be it a GPU or CPU.

For example, the image below represents a graph in TensorFlow. W, x and b are tensors over the edges of this graph. MatMul is an operation over the tensors W and x, after that Add is called and add the result of the previous operator with b.

Importing TensorFlow

To use TensorFlow, we need to import the library. We imported it and optionally gave it the name "tf", so the modules can be accessed by tf.module-name:


In [19]:
import tensorflow as tf

Building a Graph

As we said before, TensorFlow works as a graph computational model. Let's create our first graph.

To create two source operations which will output numbers we will define two constants:


In [20]:
a = tf.constant([2])
b = tf.constant([3])

After that, let's make an operation over these variables. The function tf.add() adds two elements (you could also use c = a + b).


In [21]:
c = tf.add(a,b)

Then TensorFlow needs to initialize a session to run our code. Sessions are, in a way, a context for creating a graph inside TensorFlow. Let's define our session:


In [22]:
session = tf.Session()

Let's run the session to get the result from the previous defined 'c' operation:


In [23]:
result = session.run(c)
print(result)


[5]

Close the session to release resources:


In [24]:
session.close()

To avoid having to close sessions every time, we can define them in a with block, so after running the with block the session will close automatically:


In [25]:
with tf.Session() as session:
    result = session.run(c)
    print(result)


[5]

Even this silly example of adding 2 constants to reach a simple result defines the basis of TensorFlow. Define your edge (In this case our constants), include nodes (operations, like tf.add), and start a session to build a graph.


TensorFlow Basic Elements

  • Tensor
  • Variable
  • Operation
  • Session
  • Placeholder
  • Tensorboard

What is the meaning of Tensor?

In TensorFlow all data is passed between operations in a computation graph, and these are passed in the form of Tensors, hence the name of TensorFlow.

The word __tensor__ from new latin means "that which stretches". It is a mathematical object that is named __tensor__ because an early application of tensors was the study of materials stretching under tension. The contemporary meaning of tensors can be taken as multidimensional arrays.

What are multidimensional arrays here?

Dimension Physical Representation Mathematical Object In Code
Zero Point Scalar (Single Number) [ 1 ]
One Line Vector (Series of Numbers) [ 1,2,3,4,... ]
Two Surface Matrix (Table of Numbers) [ [1,2,3,4,...], [1,2,3,4,...], [1,2,3,4,...],... ]
Three Volume Tensor (Cube of Numbers) [ [[1,2,...], [1,2,...], [1,2,...],...], [[1,2,...], [1,2,...], [1,2,...],...], [[1,2,...], [1,2,...], [1,2,...] ,...]... ]

Defining multidimensional arrays using TensorFlow

Now we will try to define such arrays using TensorFlow:


In [26]:
Scalar = tf.constant([2])
Vector = tf.constant([5,6,2])
Matrix = tf.constant([[1,2,3],[2,3,4],[3,4,5]])
Tensor = tf.constant( [ [[1,2,3],[2,3,4],[3,4,5]] , [[4,5,6],[5,6,7],[6,7,8]] , [[7,8,9],[8,9,10],[9,10,11]] ] )
with tf.Session() as session:
    result = session.run(Scalar)
    print ("Scalar (1 entry):\n %s \n" % result)
    result = session.run(Vector)
    print ("Vector (3 entries) :\n %s \n" % result)
    result = session.run(Matrix)
    print ("Matrix (3x3 entries):\n %s \n" % result)
    result = session.run(Tensor)
    print ("Tensor (3x3x3 entries) :\n %s \n" % result)


Scalar (1 entry):
 [2] 

Vector (3 entries) :
 [5 6 2] 

Matrix (3x3 entries):
 [[1 2 3]
 [2 3 4]
 [3 4 5]] 

Tensor (3x3x3 entries) :
 [[[ 1  2  3]
  [ 2  3  4]
  [ 3  4  5]]

 [[ 4  5  6]
  [ 5  6  7]
  [ 6  7  8]]

 [[ 7  8  9]
  [ 8  9 10]
  [ 9 10 11]]] 


Why Tensors?

The Tensor structure helps us by giving the freedom to shape the dataset the way we want.

And it is particularly helpful when dealing with images, due to the nature of how information in images are encoded,

Thinking about images, its easy to understand that it has a height and width, so it would make sense to represent the information contained in it with a two dimensional strucutre (a matrix)... until you remember that images have colors, and to add information about the colors, we need another dimension, and thats when Tensors become particulary helpful.

Images are encoded into color channels, the image data is represented into each color intensity in a color channel at a given point, the most common one being RGB, which means Red, Blue and Green. The information contained into an image is the intensity of each channel color into the width and height of the image, just like this:

So the intensity of the red channel at each point with width and height can be represented into a matrix, the same goes for the blue and green channels, so we end up having three matrices, and when these are combined they form a tensor.


Variables

Now that we are more familiar with the structure of data, we will take a look at how TensorFlow handles variables.

To define variables we use the command tf.variable(). To be able to use variables in a computation graph it is necessary to initialize them before running the graph in a session. This is done by running tf.global_variables_initializer().

To update the value of a variable, we simply run an assign operation that assigns a value to the variable:


In [27]:
state = tf.Variable(0)

Let's first create a simple counter, a variable that increases one unit at a time:


In [28]:
one = tf.constant(1)
new_value = tf.add(state, one)
update = tf.assign(state, new_value)

Variables must be initialized by running an initialization operation after having launched the graph. We first have to add the initialization operation to the graph:


In [29]:
init = tf.global_variables_initializer()

We then start a session to run the graph, first initialize the variables, then print the initial value of the state variable, and then run the operation of updating the state variable and printing the result after each update:


In [30]:
with tf.Session() as session:
  session.run(init)
  print(session.run(state))
  for i in range(3):
    session.run(update)
    print(session.run(state))


0
1
2
3

Placeholders

Now we know how to manipulate variables inside TensorFlow, but what about feeding data outside of a TensorFlow model?

If you want to feed data to a TensorFlow model from outside a model, you will need to use placeholders.

So what are these placeholders and what do they do?

Placeholders can be seen as "holes" in your model, "holes" which you will pass the data to, you can create them using
tf.placeholder(_datatype_), where _datatype_ specifies the type of data (integers, floating points, strings, booleans) along with its precision (8, 16, 32, 64) bits.

The definition of each data type with the respective python sintax is defined as:

Data type Python type Description
DT_FLOAT tf.float32 32 bits floating point.
DT_DOUBLE tf.float64 64 bits floating point.
DT_INT8 tf.int8 8 bits signed integer.
DT_INT16 tf.int16 16 bits signed integer.
DT_INT32 tf.int32 32 bits signed integer.
DT_INT64 tf.int64 64 bits signed integer.
DT_UINT8 tf.uint8 8 bits unsigned integer.
DT_STRING tf.string Variable length byte arrays. Each element of a Tensor is a byte array.
DT_BOOL tf.bool Boolean.
DT_COMPLEX64 tf.complex64 Complex number made of two 32 bits floating points: real and imaginary parts.
DT_COMPLEX128 tf.complex128 Complex number made of two 64 bits floating points: real and imaginary parts.
DT_QINT8 tf.qint8 8 bits signed integer used in quantized Ops.
DT_QINT32 tf.qint32 32 bits signed integer used in quantized Ops.
DT_QUINT8 tf.quint8 8 bits unsigned integer used in quantized Ops.
[[Table Source]](https://www.tensorflow.org/versions/r0.9/resources/dims_types.html)

So we create a placeholder:


In [31]:
a=tf.placeholder(tf.float32)

And define a simple multiplication operation:


In [32]:
b=a*2

Now we need to define and run the session, but since we created a "hole" in the model to pass the data, when we initialize the session we are obligated to pass an argument with the data, otherwise we would get an error.

To pass the data to the model we call the session with an extra argument feed_dict in which we should pass a dictionary with each placeholder name folowed by its respective data, just like this:


In [34]:
with tf.Session() as sess:
    result = sess.run(b,feed_dict={a:3.5})
    print (result)


7.0

Since data in TensorFlow is passed in form of multidimensional arrays we can pass any kind of tensor through the placeholders to get the answer to the simple multiplication operation:


In [35]:
dictionary={a: [ [ [1,2,3],[4,5,6],[7,8,9],[10,11,12] ] , [ [13,14,15],[16,17,18],[19,20,21],[22,23,24] ] ] }

with tf.Session() as sess:
    result = sess.run(b,feed_dict=dictionary)
    print (result)


[[[  2.   4.   6.]
  [  8.  10.  12.]
  [ 14.  16.  18.]
  [ 20.  22.  24.]]

 [[ 26.  28.  30.]
  [ 32.  34.  36.]
  [ 38.  40.  42.]
  [ 44.  46.  48.]]]

Operations

Operations are nodes that represent the mathematical operations over the tensors on a graph. These operations can be any kind of functions, like add and subtract tensor or maybe an activation function.

tf.matmul, tf.add, tf.nn.sigmoid are some of the operations in TensorFlow. These are like functions in python but operate directly over tensors and each one does a specific thing.

Other operations can be easily found in: https://www.tensorflow.org/versions/r0.9/api_docs/python/index.html

In [36]:
a = tf.constant([5])
b = tf.constant([2])
c = tf.add(a,b)
d = tf.subtract(a,b)

with tf.Session() as session:
    result = session.run(c)
    print ('c =: %s' % result)
    result = session.run(d)
    print ('d =: %s' % result)


c =: [7]
d =: [3]

Tensorboard

TensorBoard is a suite of web applications for inspecting and understanding your TensorFlow runs and graphs. TensorBoard currently supports five visualizations: scalars, images, audio, histograms, and graphs. The computations you will use in TensorFlow for things such as training a massive deep neural network, can be fairly complex and confusing, TensorBoard will make this a lot easier to understand, debug, and optimize your TensorFlow programs.

This is what a tensorboard looks like:


In [37]:
import tensorflow as tf

with tf.name_scope("Operations"):
    with tf.name_scope("Scope_a"):
        a = tf.add(1, 2, name="a")
        b = tf.multiply(a, 3, name="b")
    with tf.name_scope("Scope_b"):
        c = tf.add(4, 5, name="c")
        d = tf.multiply(c, 6, name="d")

with tf.name_scope("Scope_c"):
    e = tf.multiply(4, 5, name="e")
    f = tf.div(c, 6, name="f")
g = tf.add(b, d,  name="g")
h = tf.multiply(g, f, name="h")

with tf.Session() as sess:
    print(sess.run(h))


63

In [38]:
with tf.Session() as sess:
    writer = tf.summary.FileWriter("/home/raghav/TECH/output4", sess.graph)
    print(sess.run(h))
    writer.close()


63

Run the tensorboard with the following command on the terminal:

 tensorboard --logdir=path/to/logs/directory/